Goto

Collaborating Authors

 log null 4



Finite-Horizon Quickest Change Detection Balancing Latency with False Alarm Probability

Huang, Yu-Han, Veeravalli, Venugopal V.

arXiv.org Machine Learning

A finite-horizon variant of the quickest change detection (QCD) problem that is of relevance to learning in non-stationary environments is studied. The metric characterizing false alarms is the probability of a false alarm occurring before the horizon ends. The metric that characterizes the delay is \emph{latency}, which is the smallest value such that the probability that detection delay exceeds this value is upper bounded to a predetermined latency level. The objective is to minimize the latency (at a given latency level), while maintaining a low false alarm probability. Under the pre-specified latency and false alarm levels, a universal lower bound on the latency, which any change detection procedure needs to satisfy, is derived. Change detectors are then developed, which are order-optimal in terms of the horizon. The case where the pre- and post-change distributions are known is considered first, and then the results are generalized to the non-parametric case when they are unknown except that they are sub-Gaussian with different means. Simulations are provided to validate the theoretical results.


43207fd5e34f87c48d584fc5c11befb8-Supplemental.pdf

Neural Information Processing Systems

Is Plug-in Solver Sample Efficient for Feature-based Reinfocement Learning? DMDP, so the optimal policy exists for player 1. For this policy, neither player can benefit from change its policy alone. We give the following well-known properties of 2-TBSG without proof (see. Here we prove the three arguments in Proposition 1. 1.


A Details of the Experiments

Neural Information Processing Systems

We define δ ( s,a) = null 1 if 1{s = 0 } = 1{a = 0 }, 0 otherwise. It is straightforward to verify that this is a valid time-inhomogeneous linear MDP . The results are reported in Figure 2. As mentioned in the discussion following Theorem 4.1, it holds that These findings also shed light on the minimax optimality of the OPE problem. VA-OPE is a promising candidate for achieving minimax optimality. We further investigate this in the next subsection.


When Models Don't Collapse: On the Consistency of Iterative MLE

Barzilai, Daniel, Shamir, Ohad

arXiv.org Machine Learning

The widespread use of generative models has created a feedback loop, in which each generation of models is trained on data partially produced by its predecessors. This process has raised concerns about \emph{model collapse}: A critical degradation in performance caused by repeated training on synthetic data. However, different analyses in the literature have reached different conclusions as to the severity of model collapse. As such, it remains unclear how concerning this phenomenon is, and under which assumptions it can be avoided. To address this, we theoretically study model collapse for maximum likelihood estimation (MLE), in a natural setting where synthetic data is gradually added to the original data set. Under standard assumptions (similar to those long used for proving asymptotic consistency and normality of MLE), we establish non-asymptotic bounds showing that collapse can be avoided even as the fraction of real data vanishes. On the other hand, we prove that some assumptions (beyond MLE consistency) are indeed necessary: Without them, model collapse can occur arbitrarily quickly, even when the original data is still present in the training set. To the best of our knowledge, these are the first rigorous examples of iterative generative modeling with accumulating data that rapidly leads to model collapse.


Beyond Smoothness: Incorporating Low-Rank Analysis into Nonparametric Density Estimation

Vandermeulen, Robert A., Ledent, Antoine

arXiv.org Machine Learning

The construction and theoretical analysis of the most popular universally consistent nonparametric density estimators hinge on one functional property: smoothness. In this paper we investigate the theoretical implications of incorporating a multi-view latent variable model, a type of low-rank model, into nonparametric density estimation. To do this we perform extensive analysis on histogram-style estimators that integrate a multi-view model. Our analysis culminates in showing that there exists a universally consistent histogram-style estimator that converges to any multi-view model with a finite number of Lipschitz continuous components at a rate of $\widetilde{O}(1/\sqrt[3]{n})$ in $L^1$ error. In contrast, the standard histogram estimator can converge at a rate slower than $1/\sqrt[d]{n}$ on the same class of densities. We also introduce a new nonparametric latent variable model based on the Tucker decomposition. A rudimentary implementation of our estimators experimentally demonstrates a considerable performance improvement over the standard histogram estimator. We also provide a thorough analysis of the sample complexity of our Tucker decomposition-based model and a variety of other results. Thus, our paper provides solid theoretical foundations for extending low-rank techniques to the nonparametric setting